4 research outputs found

    Enhancing Attention’s Explanation Using Interpretable Tsetlin Machine

    Get PDF
    Explainability is one of the key factors in Natural Language Processing (NLP) specially for legal documents, medical diagnosis, and clinical text. Attention mechanism has been a popular choice for such explainability recently by estimating the relative importance of input units. Recent research has revealed, however, that such processes tend to misidentify irrelevant input units when explaining them. This is due to the fact that language representation layers are initialized by pretrained word embedding that is not context-dependent. Such a lack of context-dependent knowledge in the initial layer makes it difficult for the model to concentrate on the important aspects of input. Usually, this does not impact the performance of the model, but the explainability differs from human understanding. Hence, in this paper, we propose an ensemble method to use logic-based information from the Tsetlin Machine to embed it into the initial representation layer in the neural network to enhance the model in terms of explainability. We obtain the global clause score for each word in the vocabulary and feed it into the neural network layer as context-dependent information. Our experiments show that the ensemble method enhances the explainability of the attention layer without sacrificing any performance of the model and even outperforming in some datasets.publishedVersio

    A Lite Romanian BERT:ALR-BERT

    Get PDF
    Large-scale pre-trained language representation and its promising performance in various downstream applications have become an area of interest in the field of natural language processing (NLP). There has been huge interest in further increasing the model’s size in order to outperform the best previously obtained performances. However, at some point, increasing the model’s parameters may lead to reaching its saturation point due to the limited capacity of GPU/TPU. In addition to this, such models are mostly available in English or a shared multilingual structure. Hence, in this paper, we propose a lite BERT trained on a large corpus solely in the Romanian language, which we called “A Lite Romanian BERT (ALR-BERT)”. Based on comprehensive empirical results, ALR-BERT produces models that scale far better than the original Romanian BERT. Alongside presenting the performance on downstream tasks, we detail the analysis of the training process and its parameters. We also intend to distribute our code and model as an open source together with the downstream task.publishedVersio

    A Lite Romanian BERT:ALR-BERT

    No full text
    Large-scale pre-trained language representation and its promising performance in various downstream applications have become an area of interest in the field of natural language processing (NLP). There has been huge interest in further increasing the model’s size in order to outperform the best previously obtained performances. However, at some point, increasing the model’s parameters may lead to reaching its saturation point due to the limited capacity of GPU/TPU. In addition to this, such models are mostly available in English or a shared multilingual structure. Hence, in this paper, we propose a lite BERT trained on a large corpus solely in the Romanian language, which we called “A Lite Romanian BERT (ALR-BERT)”. Based on comprehensive empirical results, ALR-BERT produces models that scale far better than the original Romanian BERT. Alongside presenting the performance on downstream tasks, we detail the analysis of the training process and its parameters. We also intend to distribute our code and model as an open source together with the downstream task

    Research and Science Today No. 1(17)/2019

    No full text
    RESEARCH AND SCIENCE TODAY is a biannual science journal established in 2011. The journal is an informational platform that publishes assessment articles and the results of various scientific research carried out by academics. We provide the authors with the opportunity to create and/or perfect their science writing skills. Thus, each issue of the journal (two per year and at least two supplements) will contain professional articles from any academic field, authored by domestic and international academics. The goal of this journal is to pass on relevant information to undergraduate, graduate, and post-graduate students as well as to fellow academics and researchers; the topics covered are unlimited, considering its multi-disciplinary profile. Regarding the national and international visibility of Research and Science Today, it is indexed in over 30 international databases (IDB) and is present in over 200 online libraries and catalogues; therefore, anybody can easily consult the articles featured in each issue by accessing the databases or simply the website
    corecore